Integration of Speech and Gesture Inputs during Multimodal Interaction
نویسندگان
چکیده
Speech and gesture are two types of multimodal inputs that can be used to facilitate more natural humanmachine interaction in applications for which the traditional keyboard and mouse input mechanisms are inappropriate, however the possibility of their concurrent use raises the issue of how best to fuse the two inputs. This paper analyses data collected from a speech and manual gesture-based digital photo management application scenario, and from this derives assumptions and fusion thresholds with which future speech/gesture systems can be designed. Gesture input was found to overlap with speech input in nearly all multimodal constructions (95%), and was completely subsumed by speech input in most multimodal constructions (56%), in distinct contrast to previous similar analyses for combined pen and speech input, in which the pen input frequently precedes the speech input.
منابع مشابه
A Toolkit for Creating and Testing Multimodal Interface Designs
Designing and implementing applications that can handle multiple recognition-based interaction technologies such as speech and gesture inputs is a difficult task. IMBuilder and MEngine are the two components of a new toolkit for rapidly creating and testing multimodal interface designs. First, an interaction model is specified in the form of a collection of finite state machines, using a simple...
متن کاملHuman Factors and Design Issues in Multimodal (Speech/Gesture) Interface
Multimodal interfaces are the emerging technology that offers expressive, transparent, efficient, robust, and mobile human-computer interaction. In this paper, we described the speech/gesture based multimodal interface systematically from the human factors point of view. To design more practical and efficient multimodal interface, human factors issues such as user modeling, usability studies, s...
متن کاملTemporal Symbolic Integration Applied to a Multimodal System Using Gestures and Speech
This paper presents a technical approach for temporal symbol integration aimed to be generally applicable in unimodal and multimodal user interfaces. It draws its strength from symbolic data representation and an underlying rulebased system, and is embedded in a multi-agent system. The core method for temporal integration is motivated by findings from cognitive science research. We discuss its ...
متن کاملUnification-based Multimodal Integration
Recent empirical research has shown conclusive advantages of multimodal interaction over speech-only interaction for mapbased tasks. This paper describes a multimodal language processing architecture which supports interfaces allowing simultaneous input from speech and gesture recognition. Integration of spoken and gestural input is driven by uni cation of typed feature structures representing ...
متن کاملTangible Objects for the Acquisition of Multimodal Interaction Patterns
Multimodal user interfaces offer more intuitive interaction for end-users, however, usually only through predefined input schemes. This paper describes a user experiment for multimodal interaction pattern identification, using head gesture and speech inputs for a 3D graph manipulation. We show that a direct mapping between head gestures and the 3D object predominates, however even for such a si...
متن کامل